Learning a time-dependent master saliency map from eye-tracking data in videos
نویسندگان
چکیده
To predict the most salient regions of complex natural scenes, saliency models commonly compute several feature maps (contrast, orientation, motion...) and linearly combine them into a master saliency map. Since feature maps have different spatial distribution and amplitude dynamic ranges, determining their contributions to overall saliency remains an open problem. Most state-of-the-art models do not take time into account and give feature maps constant weights across the stimulus duration. However, visual exploration is a highly dynamic process shaped by many time-dependent factors. For instance, some systematic viewing patterns such as the center bias are known to dramatically vary across the time course of the exploration. In this paper, we use maximum likelihood and shrinkage methods to dynamically and jointly learn feature map and systematic viewing pattern weights directly from eye-tracking data recorded on videos. We show that these weights systematically vary as a function of time, and heavily depend upon the semantic visual category of the videos being processed. Our fusion method allows taking these variations into account, and outperforms other stateof-the-art fusion schemes using constant weights over time. The code, videos and eye-tracking data we used for this study are available online.
منابع مشابه
Compressed-Sampling-Based Image Saliency Detection in the Wavelet Domain
When watching natural scenes, an overwhelming amount of information is delivered to the Human Visual System (HVS). The optic nerve is estimated to receive around 108 bits of information a second. This large amount of information can’t be processed right away through our neural system. Visual attention mechanism enables HVS to spend neural resources efficiently, only on the selected parts of the...
متن کاملPredicting Video Saliency with Object-to-Motion CNN and Two-layer Convolutional LSTM
Over the past few years, deep neural networks (DNNs) have exhibited great success in predicting the saliency of images. However, there are few works that apply DNNs to predict the saliency of generic videos. In this paper, we propose a novel DNN-based video saliency prediction method. Specifically, we establish a large-scale eye-tracking database of videos (LEDOV), which provides sufficient dat...
متن کاملVisual saliency maps for studies of behavior of patients with neurodegenerative diseases: Observer’s versus Actor’s points of view
Finding the salient regions in videos has been a very active topic. In this work we compare the modelisation of visual attention on egocentric video recordings for two different points of view. We are interested in finding the relation between the visual saliency maps of the viewer of visual content and the actors (person executing the actions). This question is of importance because the buildi...
متن کاملDynamic Saliency Models and Human Attention: A Comparative Study on Videos
Significant progress has been made in terms of computational models of bottom-up visual attention (saliency). However, efficient ways of comparing these models for still images remain an open research question. The problem is even more challenging when dealing with videos and dynamic saliency. The paper proposes a framework for dynamicsaliency model evaluation, based on a new database of divers...
متن کاملIJSRD - International Journal for Scientific Research & Development| Vol. 5, Issue 02, 2017 | ISSN (online): 2321-0613
The spatio-temporal visual system is a computational approach to model the bottom-up visual saliency for HDR input by combining spatial and temporal visual features. The main advantage of this system is that it will reduce the cognitive processing efforts. Computational models of visual attention can be applied to areas such as computer graphics, video coding and quality assessment. The propose...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1702.00714 شماره
صفحات -
تاریخ انتشار 2017